func runtime.add

249 uses

	runtime (current package)
		alg.go#L176: 			h = typehash(a.elem, add(p, i*a.elem.size), h)
		alg.go#L185: 			h = typehash(f.typ, add(p, f.offset()), h)
		cgocall.go#L465: 			p = add(p, at.elem.size)
		cgocall.go#L491: 		p = *(*unsafe.Pointer)(add(p, goarch.PtrSize))
		cgocall.go#L514: 			p = add(p, st.elem.size)
		cgocall.go#L537: 			cgoCheckArg(f.typ, add(p, f.offset()), true, top, msg)
		cgocheck.go#L99: 		p = add(p, typ.size)
		cgocheck.go#L126: 			cgoCheckBits(add(src, -doff), datap.gcdatamask.bytedata, off+doff, size)
		cgocheck.go#L131: 			cgoCheckBits(add(src, -boff), datap.gcbssmask.bytedata, off+boff, size)
		cgocheck.go#L157: 			v := *(*unsafe.Pointer)(add(src, i))
		cgocheck.go#L175: 	src = add(src, skipBytes)
		cgocheck.go#L190: 				v := *(*unsafe.Pointer)(add(src, i))
		cgocheck.go#L232: 			src = add(src, at.elem.size)
		cgocheck.go#L250: 			src = add(src, f.typ.size)
		chan.go#L103: 		c.buf = add(unsafe.Pointer(c), hchanSize)
		chan.go#L123: 	return add(c.buf, uintptr(i)*uintptr(c.elemsize))
		checkptr.go#L42: 	end := add(ptr, size-1)
		hash64.go#L33: 		a |= uintptr(*(*byte)(add(p, s>>1))) << 8
		hash64.go#L34: 		a |= uintptr(*(*byte)(add(p, s-1))) << 16
		hash64.go#L40: 		b = r4(add(p, s-4))
		hash64.go#L46: 		b = r8(add(p, s-8))
		hash64.go#L53: 				seed = mix(r8(p)^m2, r8(add(p, 8))^seed)
		hash64.go#L54: 				seed1 = mix(r8(add(p, 16))^m3, r8(add(p, 24))^seed1)
		hash64.go#L55: 				seed2 = mix(r8(add(p, 32))^m4, r8(add(p, 40))^seed2)
		hash64.go#L56: 				p = add(p, 48)
		hash64.go#L61: 			seed = mix(r8(p)^m2, r8(add(p, 8))^seed)
		hash64.go#L62: 			p = add(p, 16)
		hash64.go#L64: 		a = r8(add(p, l-16))
		hash64.go#L65: 		b = r8(add(p, l-8))
		iface.go#L104: 		p := (**itab)(add(unsafe.Pointer(&t.entries), h*goarch.PtrSize))
		iface.go#L165: 		p := (**itab)(add(unsafe.Pointer(&t.entries), h*goarch.PtrSize))
		iface.go#L203: 	xmhdr := (*[1 << 16]method)(add(unsafe.Pointer(x), uintptr(x.moff)))[:nt:nt]
		iface.go#L356: 			x = add(x, 6)
		iface.go#L369: 			x = add(x, 4)
		iface.go#L485: 		m := *(**itab)(add(unsafe.Pointer(&t.entries), i*goarch.PtrSize))
		malloc.go#L1222: 		x = add(x, size-dataSize)
		map.go#L209: 	return *(**bmap)(add(unsafe.Pointer(b), uintptr(t.bucketsize)-goarch.PtrSize))
		map.go#L213: 	*(**bmap)(add(unsafe.Pointer(b), uintptr(t.bucketsize)-goarch.PtrSize)) = ovf
		map.go#L217: 	return add(unsafe.Pointer(b), dataOffset)
		map.go#L254: 			h.extra.nextOverflow = (*bmap)(add(unsafe.Pointer(ovf), uintptr(t.bucketsize)))
		map.go#L383: 		nextOverflow = (*bmap)(add(buckets, base*uintptr(t.bucketsize)))
		map.go#L384: 		last := (*bmap)(add(buckets, (nbuckets-1)*uintptr(t.bucketsize)))
		map.go#L419: 	b := (*bmap)(add(h.buckets, (hash&m)*uintptr(t.bucketsize)))
		map.go#L425: 		oldb := (*bmap)(add(c, (hash&m)*uintptr(t.bucketsize)))
		map.go#L440: 			k := add(unsafe.Pointer(b), dataOffset+i*uintptr(t.keysize))
		map.go#L445: 				e := add(unsafe.Pointer(b), dataOffset+bucketCnt*uintptr(t.keysize)+i*uintptr(t.elemsize))
		map.go#L480: 	b := (*bmap)(add(h.buckets, (hash&m)*uintptr(t.bucketsize)))
		map.go#L486: 		oldb := (*bmap)(add(c, (hash&m)*uintptr(t.bucketsize)))
		map.go#L501: 			k := add(unsafe.Pointer(b), dataOffset+i*uintptr(t.keysize))
		map.go#L506: 				e := add(unsafe.Pointer(b), dataOffset+bucketCnt*uintptr(t.keysize)+i*uintptr(t.elemsize))
		map.go#L524: 	b := (*bmap)(add(h.buckets, (hash&m)*uintptr(t.bucketsize)))
		map.go#L530: 		oldb := (*bmap)(add(c, (hash&m)*uintptr(t.bucketsize)))
		map.go#L545: 			k := add(unsafe.Pointer(b), dataOffset+i*uintptr(t.keysize))
		map.go#L550: 				e := add(unsafe.Pointer(b), dataOffset+bucketCnt*uintptr(t.keysize)+i*uintptr(t.elemsize))
		map.go#L612: 	b := (*bmap)(add(h.buckets, bucket*uintptr(t.bucketsize)))
		map.go#L624: 					insertk = add(unsafe.Pointer(b), dataOffset+i*uintptr(t.keysize))
		map.go#L625: 					elem = add(unsafe.Pointer(b), dataOffset+bucketCnt*uintptr(t.keysize)+i*uintptr(t.elemsize))
		map.go#L632: 			k := add(unsafe.Pointer(b), dataOffset+i*uintptr(t.keysize))
		map.go#L643: 			elem = add(unsafe.Pointer(b), dataOffset+bucketCnt*uintptr(t.keysize)+i*uintptr(t.elemsize))
		map.go#L666: 		insertk = add(unsafe.Pointer(newb), dataOffset)
		map.go#L667: 		elem = add(insertk, bucketCnt*uintptr(t.keysize))
		map.go#L728: 	b := (*bmap)(add(h.buckets, bucket*uintptr(t.bucketsize)))
		map.go#L740: 			k := add(unsafe.Pointer(b), dataOffset+i*uintptr(t.keysize))
		map.go#L754: 			e := add(unsafe.Pointer(b), dataOffset+bucketCnt*uintptr(t.keysize)+i*uintptr(t.elemsize))
		map.go#L893: 			b = (*bmap)(add(h.oldbuckets, oldbucket*uintptr(t.bucketsize)))
		map.go#L897: 				b = (*bmap)(add(it.buckets, bucket*uintptr(t.bucketsize)))
		map.go#L901: 			b = (*bmap)(add(it.buckets, bucket*uintptr(t.bucketsize)))
		map.go#L918: 		k := add(unsafe.Pointer(b), dataOffset+uintptr(offi)*uintptr(t.keysize))
		map.go#L922: 		e := add(unsafe.Pointer(b), dataOffset+bucketCnt*uintptr(t.keysize)+uintptr(offi)*uintptr(t.elemsize))
		map.go#L1138: 	b := (*bmap)(add(h.oldbuckets, bucket*uintptr(t.bucketsize)))
		map.go#L1151: 	b := (*bmap)(add(h.oldbuckets, oldbucket*uintptr(t.bucketsize)))
		map.go#L1160: 		x.b = (*bmap)(add(h.buckets, oldbucket*uintptr(t.bucketsize)))
		map.go#L1161: 		x.k = add(unsafe.Pointer(x.b), dataOffset)
		map.go#L1162: 		x.e = add(x.k, bucketCnt*uintptr(t.keysize))
		map.go#L1168: 			y.b = (*bmap)(add(h.buckets, (oldbucket+newbit)*uintptr(t.bucketsize)))
		map.go#L1169: 			y.k = add(unsafe.Pointer(y.b), dataOffset)
		map.go#L1170: 			y.e = add(y.k, bucketCnt*uintptr(t.keysize))
		map.go#L1174: 			k := add(unsafe.Pointer(b), dataOffset)
		map.go#L1175: 			e := add(k, bucketCnt*uintptr(t.keysize))
		map.go#L1176: 			for i := 0; i < bucketCnt; i, k, e = i+1, add(k, uintptr(t.keysize)), add(e, uintptr(t.elemsize)) {
		map.go#L1225: 					dst.k = add(unsafe.Pointer(dst.b), dataOffset)
		map.go#L1226: 					dst.e = add(dst.k, bucketCnt*uintptr(t.keysize))
		map.go#L1244: 				dst.k = add(dst.k, uintptr(t.keysize))
		map.go#L1245: 				dst.e = add(dst.e, uintptr(t.elemsize))
		map.go#L1250: 			b := add(h.oldbuckets, oldbucket*uintptr(t.bucketsize))
		map.go#L1253: 			ptr := add(b, dataOffset)
		map_fast32.go#L31: 		b = (*bmap)(add(h.buckets, (hash&m)*uintptr(t.bucketsize)))
		map_fast32.go#L37: 			oldb := (*bmap)(add(c, (hash&m)*uintptr(t.bucketsize)))
		map_fast32.go#L44: 		for i, k := uintptr(0), b.keys(); i < bucketCnt; i, k = i+1, add(k, 4) {
		map_fast32.go#L46: 				return add(unsafe.Pointer(b), dataOffset+bucketCnt*4+i*uintptr(t.elemsize))
		map_fast32.go#L71: 		b = (*bmap)(add(h.buckets, (hash&m)*uintptr(t.bucketsize)))
		map_fast32.go#L77: 			oldb := (*bmap)(add(c, (hash&m)*uintptr(t.bucketsize)))
		map_fast32.go#L84: 		for i, k := uintptr(0), b.keys(); i < bucketCnt; i, k = i+1, add(k, 4) {
		map_fast32.go#L86: 				return add(unsafe.Pointer(b), dataOffset+bucketCnt*4+i*uintptr(t.elemsize)), true
		map_fast32.go#L118: 	b := (*bmap)(add(h.buckets, bucket*uintptr(t.bucketsize)))
		map_fast32.go#L137: 			k := *((*uint32)(add(unsafe.Pointer(b), dataOffset+i*4)))
		map_fast32.go#L168: 	insertk = add(unsafe.Pointer(insertb), dataOffset+inserti*4)
		map_fast32.go#L175: 	elem := add(unsafe.Pointer(insertb), dataOffset+bucketCnt*4+inserti*uintptr(t.elemsize))
		map_fast32.go#L208: 	b := (*bmap)(add(h.buckets, bucket*uintptr(t.bucketsize)))
		map_fast32.go#L227: 			k := *((*unsafe.Pointer)(add(unsafe.Pointer(b), dataOffset+i*4)))
		map_fast32.go#L258: 	insertk = add(unsafe.Pointer(insertb), dataOffset+inserti*4)
		map_fast32.go#L265: 	elem := add(unsafe.Pointer(insertb), dataOffset+bucketCnt*4+inserti*uintptr(t.elemsize))
		map_fast32.go#L294: 	b := (*bmap)(add(h.buckets, bucket*uintptr(t.bucketsize)))
		map_fast32.go#L298: 		for i, k := uintptr(0), b.keys(); i < bucketCnt; i, k = i+1, add(k, 4) {
		map_fast32.go#L310: 			e := add(unsafe.Pointer(b), dataOffset+bucketCnt*4+i*uintptr(t.elemsize))
		map_fast32.go#L375: 	b := (*bmap)(add(h.oldbuckets, oldbucket*uintptr(t.bucketsize)))
		map_fast32.go#L384: 		x.b = (*bmap)(add(h.buckets, oldbucket*uintptr(t.bucketsize)))
		map_fast32.go#L385: 		x.k = add(unsafe.Pointer(x.b), dataOffset)
		map_fast32.go#L386: 		x.e = add(x.k, bucketCnt*4)
		map_fast32.go#L392: 			y.b = (*bmap)(add(h.buckets, (oldbucket+newbit)*uintptr(t.bucketsize)))
		map_fast32.go#L393: 			y.k = add(unsafe.Pointer(y.b), dataOffset)
		map_fast32.go#L394: 			y.e = add(y.k, bucketCnt*4)
		map_fast32.go#L398: 			k := add(unsafe.Pointer(b), dataOffset)
		map_fast32.go#L399: 			e := add(k, bucketCnt*4)
		map_fast32.go#L400: 			for i := 0; i < bucketCnt; i, k, e = i+1, add(k, 4), add(e, uintptr(t.elemsize)) {
		map_fast32.go#L425: 					dst.k = add(unsafe.Pointer(dst.b), dataOffset)
		map_fast32.go#L426: 					dst.e = add(dst.k, bucketCnt*4)
		map_fast32.go#L444: 				dst.k = add(dst.k, 4)
		map_fast32.go#L445: 				dst.e = add(dst.e, uintptr(t.elemsize))
		map_fast32.go#L450: 			b := add(h.oldbuckets, oldbucket*uintptr(t.bucketsize))
		map_fast32.go#L453: 			ptr := add(b, dataOffset)
		map_fast64.go#L31: 		b = (*bmap)(add(h.buckets, (hash&m)*uintptr(t.bucketsize)))
		map_fast64.go#L37: 			oldb := (*bmap)(add(c, (hash&m)*uintptr(t.bucketsize)))
		map_fast64.go#L44: 		for i, k := uintptr(0), b.keys(); i < bucketCnt; i, k = i+1, add(k, 8) {
		map_fast64.go#L46: 				return add(unsafe.Pointer(b), dataOffset+bucketCnt*8+i*uintptr(t.elemsize))
		map_fast64.go#L71: 		b = (*bmap)(add(h.buckets, (hash&m)*uintptr(t.bucketsize)))
		map_fast64.go#L77: 			oldb := (*bmap)(add(c, (hash&m)*uintptr(t.bucketsize)))
		map_fast64.go#L84: 		for i, k := uintptr(0), b.keys(); i < bucketCnt; i, k = i+1, add(k, 8) {
		map_fast64.go#L86: 				return add(unsafe.Pointer(b), dataOffset+bucketCnt*8+i*uintptr(t.elemsize)), true
		map_fast64.go#L118: 	b := (*bmap)(add(h.buckets, bucket*uintptr(t.bucketsize)))
		map_fast64.go#L137: 			k := *((*uint64)(add(unsafe.Pointer(b), dataOffset+i*8)))
		map_fast64.go#L168: 	insertk = add(unsafe.Pointer(insertb), dataOffset+inserti*8)
		map_fast64.go#L175: 	elem := add(unsafe.Pointer(insertb), dataOffset+bucketCnt*8+inserti*uintptr(t.elemsize))
		map_fast64.go#L208: 	b := (*bmap)(add(h.buckets, bucket*uintptr(t.bucketsize)))
		map_fast64.go#L227: 			k := *((*unsafe.Pointer)(add(unsafe.Pointer(b), dataOffset+i*8)))
		map_fast64.go#L258: 	insertk = add(unsafe.Pointer(insertb), dataOffset+inserti*8)
		map_fast64.go#L265: 	elem := add(unsafe.Pointer(insertb), dataOffset+bucketCnt*8+inserti*uintptr(t.elemsize))
		map_fast64.go#L294: 	b := (*bmap)(add(h.buckets, bucket*uintptr(t.bucketsize)))
		map_fast64.go#L298: 		for i, k := uintptr(0), b.keys(); i < bucketCnt; i, k = i+1, add(k, 8) {
		map_fast64.go#L312: 			e := add(unsafe.Pointer(b), dataOffset+bucketCnt*8+i*uintptr(t.elemsize))
		map_fast64.go#L377: 	b := (*bmap)(add(h.oldbuckets, oldbucket*uintptr(t.bucketsize)))
		map_fast64.go#L386: 		x.b = (*bmap)(add(h.buckets, oldbucket*uintptr(t.bucketsize)))
		map_fast64.go#L387: 		x.k = add(unsafe.Pointer(x.b), dataOffset)
		map_fast64.go#L388: 		x.e = add(x.k, bucketCnt*8)
		map_fast64.go#L394: 			y.b = (*bmap)(add(h.buckets, (oldbucket+newbit)*uintptr(t.bucketsize)))
		map_fast64.go#L395: 			y.k = add(unsafe.Pointer(y.b), dataOffset)
		map_fast64.go#L396: 			y.e = add(y.k, bucketCnt*8)
		map_fast64.go#L400: 			k := add(unsafe.Pointer(b), dataOffset)
		map_fast64.go#L401: 			e := add(k, bucketCnt*8)
		map_fast64.go#L402: 			for i := 0; i < bucketCnt; i, k, e = i+1, add(k, 8), add(e, uintptr(t.elemsize)) {
		map_fast64.go#L427: 					dst.k = add(unsafe.Pointer(dst.b), dataOffset)
		map_fast64.go#L428: 					dst.e = add(dst.k, bucketCnt*8)
		map_fast64.go#L452: 				dst.k = add(dst.k, 8)
		map_fast64.go#L453: 				dst.e = add(dst.e, uintptr(t.elemsize))
		map_fast64.go#L458: 			b := add(h.oldbuckets, oldbucket*uintptr(t.bucketsize))
		map_fast64.go#L461: 			ptr := add(b, dataOffset)
		map_faststr.go#L30: 			for i, kptr := uintptr(0), b.keys(); i < bucketCnt; i, kptr = i+1, add(kptr, 2*goarch.PtrSize) {
		map_faststr.go#L39: 					return add(unsafe.Pointer(b), dataOffset+bucketCnt*2*goarch.PtrSize+i*uintptr(t.elemsize))
		map_faststr.go#L46: 		for i, kptr := uintptr(0), b.keys(); i < bucketCnt; i, kptr = i+1, add(kptr, 2*goarch.PtrSize) {
		map_faststr.go#L55: 				return add(unsafe.Pointer(b), dataOffset+bucketCnt*2*goarch.PtrSize+i*uintptr(t.elemsize))
		map_faststr.go#L62: 			if *((*[4]byte)(add(key.str, uintptr(key.len)-4))) != *((*[4]byte)(add(k.str, uintptr(key.len)-4))) {
		map_faststr.go#L72: 			k := (*stringStruct)(add(unsafe.Pointer(b), dataOffset+keymaybe*2*goarch.PtrSize))
		map_faststr.go#L74: 				return add(unsafe.Pointer(b), dataOffset+bucketCnt*2*goarch.PtrSize+keymaybe*uintptr(t.elemsize))
		map_faststr.go#L82: 	b := (*bmap)(add(h.buckets, (hash&m)*uintptr(t.bucketsize)))
		map_faststr.go#L88: 		oldb := (*bmap)(add(c, (hash&m)*uintptr(t.bucketsize)))
		map_faststr.go#L95: 		for i, kptr := uintptr(0), b.keys(); i < bucketCnt; i, kptr = i+1, add(kptr, 2*goarch.PtrSize) {
		map_faststr.go#L101: 				return add(unsafe.Pointer(b), dataOffset+bucketCnt*2*goarch.PtrSize+i*uintptr(t.elemsize))
		map_faststr.go#L125: 			for i, kptr := uintptr(0), b.keys(); i < bucketCnt; i, kptr = i+1, add(kptr, 2*goarch.PtrSize) {
		map_faststr.go#L134: 					return add(unsafe.Pointer(b), dataOffset+bucketCnt*2*goarch.PtrSize+i*uintptr(t.elemsize)), true
		map_faststr.go#L141: 		for i, kptr := uintptr(0), b.keys(); i < bucketCnt; i, kptr = i+1, add(kptr, 2*goarch.PtrSize) {
		map_faststr.go#L150: 				return add(unsafe.Pointer(b), dataOffset+bucketCnt*2*goarch.PtrSize+i*uintptr(t.elemsize)), true
		map_faststr.go#L157: 			if *((*[4]byte)(add(key.str, uintptr(key.len)-4))) != *((*[4]byte)(add(k.str, uintptr(key.len)-4))) {
		map_faststr.go#L167: 			k := (*stringStruct)(add(unsafe.Pointer(b), dataOffset+keymaybe*2*goarch.PtrSize))
		map_faststr.go#L169: 				return add(unsafe.Pointer(b), dataOffset+bucketCnt*2*goarch.PtrSize+keymaybe*uintptr(t.elemsize)), true
		map_faststr.go#L177: 	b := (*bmap)(add(h.buckets, (hash&m)*uintptr(t.bucketsize)))
		map_faststr.go#L183: 		oldb := (*bmap)(add(c, (hash&m)*uintptr(t.bucketsize)))
		map_faststr.go#L190: 		for i, kptr := uintptr(0), b.keys(); i < bucketCnt; i, kptr = i+1, add(kptr, 2*goarch.PtrSize) {
		map_faststr.go#L196: 				return add(unsafe.Pointer(b), dataOffset+bucketCnt*2*goarch.PtrSize+i*uintptr(t.elemsize)), true
		map_faststr.go#L229: 	b := (*bmap)(add(h.buckets, bucket*uintptr(t.bucketsize)))
		map_faststr.go#L249: 			k := (*stringStruct)(add(unsafe.Pointer(b), dataOffset+i*2*goarch.PtrSize))
		map_faststr.go#L287: 	insertk = add(unsafe.Pointer(insertb), dataOffset+inserti*2*goarch.PtrSize)
		map_faststr.go#L293: 	elem := add(unsafe.Pointer(insertb), dataOffset+bucketCnt*2*goarch.PtrSize+inserti*uintptr(t.elemsize))
		map_faststr.go#L323: 	b := (*bmap)(add(h.buckets, bucket*uintptr(t.bucketsize)))
		map_faststr.go#L328: 		for i, kptr := uintptr(0), b.keys(); i < bucketCnt; i, kptr = i+1, add(kptr, 2*goarch.PtrSize) {
		map_faststr.go#L338: 			e := add(unsafe.Pointer(b), dataOffset+bucketCnt*2*goarch.PtrSize+i*uintptr(t.elemsize))
		map_faststr.go#L403: 	b := (*bmap)(add(h.oldbuckets, oldbucket*uintptr(t.bucketsize)))
		map_faststr.go#L412: 		x.b = (*bmap)(add(h.buckets, oldbucket*uintptr(t.bucketsize)))
		map_faststr.go#L413: 		x.k = add(unsafe.Pointer(x.b), dataOffset)
		map_faststr.go#L414: 		x.e = add(x.k, bucketCnt*2*goarch.PtrSize)
		map_faststr.go#L420: 			y.b = (*bmap)(add(h.buckets, (oldbucket+newbit)*uintptr(t.bucketsize)))
		map_faststr.go#L421: 			y.k = add(unsafe.Pointer(y.b), dataOffset)
		map_faststr.go#L422: 			y.e = add(y.k, bucketCnt*2*goarch.PtrSize)
		map_faststr.go#L426: 			k := add(unsafe.Pointer(b), dataOffset)
		map_faststr.go#L427: 			e := add(k, bucketCnt*2*goarch.PtrSize)
		map_faststr.go#L428: 			for i := 0; i < bucketCnt; i, k, e = i+1, add(k, 2*goarch.PtrSize), add(e, uintptr(t.elemsize)) {
		map_faststr.go#L453: 					dst.k = add(unsafe.Pointer(dst.b), dataOffset)
		map_faststr.go#L454: 					dst.e = add(dst.k, bucketCnt*2*goarch.PtrSize)
		map_faststr.go#L467: 				dst.k = add(dst.k, 2*goarch.PtrSize)
		map_faststr.go#L468: 				dst.e = add(dst.e, uintptr(t.elemsize))
		map_faststr.go#L473: 			b := add(h.oldbuckets, oldbucket*uintptr(t.bucketsize))
		map_faststr.go#L476: 			ptr := add(b, dataOffset)
		mgcmark.go#L281: 	ptrmask := (*uint8)(add(unsafe.Pointer(ptrmask0), uintptr(shard)*(rootBlockBytes/(8*goarch.PtrSize))))
		mgcsweep.go#L833: 		*(*uint32)(add(x, i)) = 0xdeadbeef
		mpagealloc_64bit.go#L119: 			offAddr{uintptr(add(base, baseOffset))},
		mpagealloc_64bit.go#L120: 			offAddr{uintptr(add(base, limitOffset))},
		mprof.go#L183: 	stk := (*[maxStack]uintptr)(add(unsafe.Pointer(b), unsafe.Sizeof(*b)))
		mprof.go#L192: 	data := add(unsafe.Pointer(b), unsafe.Sizeof(*b)+b.nstk*unsafe.Sizeof(uintptr(0)))
		mprof.go#L201: 	data := add(unsafe.Pointer(b), unsafe.Sizeof(*b)+b.nstk*unsafe.Sizeof(uintptr(0)))
		mspanset.go#L85: 		blockp := add(spine, goarch.PtrSize*top)
		mspanset.go#L127: 		blockp := add(b.spine, goarch.PtrSize*top)
		mspanset.go#L184: 	blockp := add(spine, goarch.PtrSize*uintptr(top))
		mspanset.go#L244: 		blockp := (**spanSetBlock)(add(b.spine, goarch.PtrSize*uintptr(top)))
		netpoll.go#L625: 			pd := (*pollDesc)(add(mem, i*pdSize))
		os_linux.go#L230: 	auxv := (*[1 << 28]uintptr)(add(unsafe.Pointer(argv), uintptr(n)*goarch.PtrSize))
		panic.go#L685: 		fd = add(fd, unsafe.Sizeof(b))
		proc.go#L568: 	return *(**g)(add(unsafe.Pointer(ptr), i*goarch.PtrSize))
		proc.go#L6197: 			p := add(unsafe.Pointer(t), (3+i)*goarch.PtrSize)
		proc.go#L6218: 		firstFunc := add(unsafe.Pointer(t), (3+t.ndeps)*goarch.PtrSize)
		proc.go#L6220: 			p := add(firstFunc, i*goarch.PtrSize)
		runtime1.go#L58: 	return *(**byte)(add(unsafe.Pointer(argv), uintptr(i)*goarch.PtrSize))
		signal_linux_amd64.go#L55: 	*(*uintptr)(add(unsafe.Pointer(c.info), 2*goarch.PtrSize)) = uintptr(x)
		slice.go#L59: 			memclrNoHeapPointers(add(to, copymem), tomem-copymem)
		slice.go#L275: 		memclrNoHeapPointers(add(p, newlenmem), capmem-newlenmem)
		stack.go#L613: 				print("        ", add(scanp, (i+j)*goarch.PtrSize), ":", ptrnames[bv.ptrbit(i+j)], ":", hex(*(*uintptr)(add(scanp, (i+j)*goarch.PtrSize))), " # ", i, " ", *addb(bv.bytedata, i/8), "\n")
		stack.go#L620: 			pp := (*uintptr)(add(scanp, (i+j)*goarch.PtrSize))
		stack.go#L1345: 			p = add(p, goarch.PtrSize)
		string.go#L103: 			p = add(p, 7)
		string.go#L288: 		memclrNoHeapPointers(add(p, uintptr(size)), cap-uintptr(size))
		string.go#L303: 		memclrNoHeapPointers(add(p, uintptr(size)*4), mem-uintptr(size)*4)
		stubs.go#L17: func add(p unsafe.Pointer, x uintptr) unsafe.Pointer {
		symtab.go#L841: 	ffb := (*findfuncbucket)(add(unsafe.Pointer(datap.findfunctab), b*unsafe.Sizeof(findfuncbucket{})))
		symtab.go#L1066: 	return *(*uint32)(add(unsafe.Pointer(&f.nfuncdata), unsafe.Sizeof(f.nfuncdata)+uintptr(table)*4))
		traceback.go#L595: 		bits := *(*uint8)(add(liveInfo, uintptr(liveIdx)+uintptr(slotIdx/8)))
		traceback.go#L600: 		x := readUnaligned64(add(argp, uintptr(off)))
		type.go#L301: 	return (*[1 << 20]*_type)(add(unsafe.Pointer(t), uadd))[:t.inCount]
		type.go#L311: 	return (*[1 << 20]*_type)(add(unsafe.Pointer(t), uadd))[t.inCount : t.inCount+outCount]
		type.go#L431: 	return (*byte)(add(unsafe.Pointer(n.bytes), uintptr(off)))
		vdso_linux.go#L113: 		pt := (*elfPhdr)(add(pt, uintptr(i)*unsafe.Sizeof(elfPhdr{})))
		vdso_linux.go#L192: 			aux := (*elfVerdaux)(add(unsafe.Pointer(def), uintptr(def.vd_aux)))
		vdso_linux.go#L201: 		def = (*elfVerdef)(add(unsafe.Pointer(def), uintptr(def.vd_next)))